Goto

Collaborating Authors

 rectangular matrix


Higher-Order Singular-Value Derivatives of Rectangular Real Matrices

Luo, Róisín, McDermott, James, O'Riordan, Colm

arXiv.org Machine Learning

We present a theoretical framework for deriving the general $n$-th order Fréchet derivatives of singular values in real rectangular matrices, by leveraging reduced resolvent operators from Kato's analytic perturbation theory for self-adjoint operators. Deriving closed-form expressions for higher-order derivatives of singular values is notoriously challenging through standard matrix-analysis techniques. To overcome this, we treat a real rectangular matrix as a compact operator on a finite-dimensional Hilbert space, and embed the rectangular matrix into a block self-adjoint operator so that non-symmetric perturbations are captured. Applying Kato's asymptotic eigenvalue expansion to this construction, we obtain a general, closed-form expression for the infinitesimal $n$-th order spectral variations. Specializing to $n=2$ and deploying on a Kronecker-product representation with matrix convention yield the Hessian of a singular value, not found in literature. By bridging abstract operator-theoretic perturbation theory with matrices, our framework equips researchers with a practical toolkit for higher-order spectral sensitivity studies in random matrix applications (e.g., adversarial perturbation in deep learning).


Matrix Completion from General Deterministic Sampling Patterns

Lee, Hanbyul, Mazumder, Rahul, Song, Qifan, Honorio, Jean

arXiv.org Artificial Intelligence

Low-rank matrix completion is to exactly or approximately recover an underlying rank-r matrix from a small number of observed entries of the matrix. It has received much attention in a wide range of applications including collaborative filtering [Goldberg et al., 1992], phase retrieval [Candes et al., 2015] and image processing [Chen and Suter, 2004]. In research on establishing provable guarantees for low-rank matrix completion methods, typical assumptions considered are as follows: first, the underlying matrix is incoherent; second, observable matrix entries are sampled according to a probabilistic (usually uniform) model. However, the latter assumption is easily violated in numerous situations; it is unlikely realistic for the sampling patterns to be uniformly at random outside experimental settings, and it may not be even reasonable to model the sampling patterns as random. With this motivation, we aim to tackle the low-rank matrix completion problem without imposing any model assumptions on the sampling patterns. Even though there have been several works on non-random sampling schemes, they have imposed additional structural assumptions on the sampling pattern which are also not applicable to many real-world scenarios. For example, Heiman et al. [2014], Bhojanapalli and Jain [2014] and Burnwal and Vidyasagar [2020] assumed that the number of observed entries is the same for each row and column of the matrix, and Bishop and Yu [2014] introduced systematic assumptions on the subsets of the observed entries.


Detection problems in the spiked matrix models

Jung, Ji Hyung, Chung, Hye Won, Lee, Ji Oon

arXiv.org Artificial Intelligence

We study the statistical decision process of detecting the low-rank signal from various signal-plus-noise type data matrices, known as the spiked random matrix models. We first show that the principal component analysis can be improved by entrywise pre-transforming the data matrix if the noise is non-Gaussian, generalizing the known results for the spiked random matrix models with rank-1 signals. As an intermediate step, we find out sharp phase transition thresholds for the extreme eigenvalues of spiked random matrices, which generalize the Baik-Ben Arous-P\'{e}ch\'{e} (BBP) transition. We also prove the central limit theorem for the linear spectral statistics for the spiked random matrices and propose a hypothesis test based on it, which does not depend on the distribution of the signal or the noise. When the noise is non-Gaussian noise, the test can be improved with an entrywise transformation to the data matrix with additive noise. We also introduce an algorithm that estimates the rank of the signal when it is not known a priori.


Detection of Signal in the Spiked Rectangular Models

Jung, Ji Hyung, Chung, Hye Won, Lee, Ji Oon

arXiv.org Machine Learning

We consider the problem of detecting signals in the rank-one signal-plus-noise data matrix models that generalize the spiked Wishart matrices. We show that the principal component analysis can be improved by pre-transforming the matrix entries if the noise is non-Gaussian. As an intermediate step, we prove a sharp phase transition of the largest eigenvalues of spiked rectangular matrices, which extends the Baik-Ben Arous-P\'ech\'e (BBP) transition. We also propose a hypothesis test to detect the presence of signal with low computational complexity, based on the linear spectral statistics, which minimizes the sum of the Type-I and Type-II errors when the noise is Gaussian.

  eigenvalue, matrix, rectangular matrix, (14 more...)
2104.13517
  Country:
  Genre: Research Report (0.64)
  Industry: Government > Regional Government (0.46)

Linear Algebra 101 -- Part 9: Singular Value Decomposition (SVD)

#artificialintelligence

Singular Value Decomposition (SVD) is another type of decomposition. Unlike eigendecomposition where the matrix you want to decompose has to be a square matrix, SVD allows you to decompose a rectangular matrix (a matrix that has different numbers of rows and columns). This is often more useful in a real-life scenario since the rectangular matrix could represent a wide variety of data that's not a square matrix. First, let's look at the definition itself. As you can see, SVD decomposes the matrix into 3 different matrices.


A Gentle Introduction to Singular-Value Decomposition for Machine Learning - Machine Learning Mastery

@machinelearnbot

The diagonal values in the Sigma matrix are known as the singular values of the original matrix A. The columns of the U matrix are called the left-singular vectors of A, and the columns of V are called the right-singular vectors of A. The SVD is calculated via iterative numerical methods. We will not go into the details of these methods. Every rectangular matrix has a singular value decomposition, although the resulting matrices may contain complex numbers and the limitations of floating point arithmetic may cause some matrices to fail to decompose neatly. The singular value decomposition (SVD) provides another way to factorize a matrix, into singular vectors and singular values. The SVD allows us to discover some of the same kind of information as the eigendecomposition. However, the SVD is more generally applicable.


VB calibration to improve the interface between phone recognizer and i-vector extractor

Brümmer, Niko

arXiv.org Machine Learning

The EM training algorithm of the classical i-vector extractor is often incorrectly described as a maximum-likelihood method. The i-vector model is however intractable: the likelihood itself and the hidden-variable posteriors needed for the EM algorithm cannot be computed in closed form. We show here that the classical i-vector extractor recipe is actually a mean-field variational Bayes (VB) recipe. This theoretical VB interpretation turns out to be of further use, because it also offers an interpretation of the newer phonetic i-vector extractor recipe, thereby unifying the two flavours of extractor. More importantly, the VB interpretation is also practically useful: it suggests ways of modifying existing i-vector extractors to make them more accurate. In particular, in existing methods, the approximate VB posterior for the GMM states is fixed, while only the parameters of the generative model are adapted. Here we explore the possibility of also mildly adjusting (calibrating) those posteriors, so that they better fit the generative model.